bleu2vec: the Painfully Familiar Metric on Continuous Vector Space Steroids

نویسندگان

  • Andre Tättar
  • Mark Fishel
چکیده

In this participation in the WMT’2017 metrics shared task we implement a fuzzy match score for n-gram precisions in the BLEU metric. To do this we learn ngram embeddings; we describe two ways of extending the WORD2VEC approach to do so. Evaluation results show that the introduced score beats the original BLEU metric on system and segment level. 1 The Painfully Familiar Metric The BLEU metric (Papineni et al., 2002) has deeply rooted in the machine translation community and is used in virtually every paper on machine translation methods. Despite the wellknown criticism (Callison-Burch et al., 2006) and a decade of collective efforts to come up with a better translation quality metric (from CallisonBurch et al., 2007 to Bojar et al., 2016) it still appeals with its ease of implementation, language independence and competitive agreement rate with human judgments, with the only viable alternative on all three accounts being the recently introduced CHRF (Popovic, 2015). The original version of BLEU is harsh on single sentences: one of the factors of the score is a geometric mean of n-gram precisions between the translation hypothesis and reference(s) and as a result sentences without 4-gram matches get a score of 0, even if there are good unigram, bigram and possibly trigram matches. There have been several attempts to “soften” this approach by using arithmetic mean instead (NIST, Doddington, 2002), allowing for partial matches using lemmatization and synonyms (METEOR, Banerjee and Lavie, 2005) and directly implementing fuzzy matches between n-grams (LEBLEU, Virpioja and Grönroos, 2015). Our work is most closely related to LEBLEU, where BLEU is augmented with fuzzy matches based on the character-level Levenshtein distance. Here we use independently learned word and n-gram embeddings instead. 2 The Continuous Vector Space Steroids Together with neural networks came the necessity to map sparse discrete values (like natural language words) into dense continuous vector representations. This is done explicitly e.g. with WORD2VEC (Mikolov et al., 2013), as well as learned as part of the whole learning process in neural networks-based language models (Mikolov et al., 2010) and translation approaches (Bahdanau et al., 2015). The approach of learning embeddings has since been extended for example to items in a relational database (Barkan and Koenigstein, 2016), sentences and documents (Le and Mikolov, 2014) and even users (Amir et al., 2017). The core part of this work consists of n-gram embeddings, the aim of which is to find similarities between short phrases like “research paper” and “scientific article”, or “do not like” and “hate”. We propose two solutions, both reducing the problem to the original WORD2VEC ; the first one only handles n-grams of the same length while the second one is more general. These are described in the following sections. 2.1 Separate N-gram Embeddings Our first approach is learning separate embedding models for unigrams, bigrams and trigrams. While

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Some properties of continuous linear operators in topological vector PN-spaces

The notion of a probabilistic metric  space  corresponds to thesituations when we do not know exactly the  distance.  Probabilistic Metric space was  introduced by Karl Menger. Alsina, Schweizer and Sklar gave a general definition of  probabilistic normed space based on the definition of Menger [1]. In this note we study the PN spaces which are  topological vector spaces and the open mapping an...

متن کامل

POINT DERIVATIONS ON BANACH ALGEBRAS OF α-LIPSCHITZ VECTOR-VALUED OPERATORS

The Lipschitz function algebras were first defined in the 1960s by some mathematicians, including Schubert. Initially, the Lipschitz real-value and complex-value functions are defined and quantitative properties of these algebras are investigated. Over time these algebras have been studied and generalized by many mathematicians such as Cao, Zhang, Xu, Weaver, and others. Let  be a non-emp...

متن کامل

Common fixed point results on vector metric spaces

In this paper we consider the so called a vector metric space, which is a generalization of metric space, where the metric is Riesz space valued. We prove some common fixed point theorems for three mappings in this space. Obtained results extend and generalize well-known comparable results in the literature.

متن کامل

New best proximity point results in G-metric space

Best approximation results provide an approximate solution to the fixed point equation $Tx=x$, when the non-self mapping $T$ has no fixed point. In particular, a well-known best approximation theorem, due to Fan cite{5}, asserts that if $K$ is a nonempty compact convex subset of a Hausdorff locally convex topological vector space $E$ and $T:Krightarrow E$ is a continuous mapping, then there exi...

متن کامل

On the topological equivalence of some generalized metric spaces

‎The aim of this paper is to establish the equivalence between the concepts‎ ‎of an $S$-metric space and a cone $S$-metric space using some topological‎ ‎approaches‎. ‎We introduce a new notion of a $TVS$-cone $S$-metric space using‎ ‎some facts about topological vector spaces‎. ‎We see that the known results on‎ ‎cone $S$-metric spaces (or $N$-cone metric spaces) can be directly obtained‎ from...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2017